Goto

Collaborating Authors

 feature selection algorithm



AlgorithmicStabilityandGeneralizationofan UnsupervisedFeatureSelectionAlgorithm

Neural Information Processing Systems

Algorithmic stability is a key characteristic of an algorithm regarding its sensitivity to perturbations of input samples. In this paper,we propose an innovativeunsupervised feature selection algorithm attaining this stability with provable guarantees.


HeFS: Helper-Enhanced Feature Selection via Pareto-Optimized Genetic Search

Fan, Yusi, Wang, Tian, Yan, Zhiying, Liu, Chang, Zhou, Qiong, Lu, Qi, Guo, Zhehao, Deng, Ziqi, Zhu, Wenyu, Zhang, Ruochi, Zhou, Fengfeng

arXiv.org Artificial Intelligence

Feature selection is a combinatorial optimization problem that is NP -hard. Conventional approaches often employ heuristic or greedy strategies, which are prone to premature convergence and may fail to capture subtle yet informative features. This limitation becomes especially critical in high - dimensional datasets, where complex and interdependent feature relationships prevail. We introduce the HeFS (Helper - Enhanced Feature Selection) framework to refine feature subsets produced by existing algorithms. HeFS systematically searches the residual feature space to identify a Helper Set-- features that complement the original subset and improve classification performance. The approach employs a biased initialization scheme and a ratio-guided mutation mechanism within a genetic algorithm, coupled with Pareto - based multi - objective optimization to jointly maximize predictive accuracy and feature complementarity. Experiments on 18 benchmark datasets demonstrate that HeFS consistently identifies overlooked yet informative features and achieves superior performance over state-of-the - art methods, including in challenging domains such as gastric cancer classification, drug toxicity prediction, and computer science applications.



Export Reviews, Discussions, Author Feedback and Meta-Reviews

Neural Information Processing Systems

"NIPS Neural Information Processing Systems 8-11th December 2014, Montreal, Canada",,, "Paper ID:","1871" "Title:","Parallel Feature Selection Inspired by Group Testing" Current Reviews First provide a summary of the paper, and then address the following criteria: Quality, clarity, originality and significance. In this paper a novel and interesting parallel feature selection framework based on group testing is proposed for large scale data. As the author claimed, the presented method can speed up the feature selection algorithm and provide superior performance than other existing methods especially on very high dimensional dataset. The proposed framework for parallel feature selection is well defined with sufficient theoretical analysis. The author has proved that KL divergence and MI is C-separable under certain conditions.



ADSEL: Adaptive dual self-expression learning for EEG feature selection via incomplete multi-dimensional emotional tagging

Yu, Tianze, Zhang, Junming, Dong, Wenjia, Xu, Xueyuan, Zhuo, Li

arXiv.org Artificial Intelligence

EEG based multi-dimension emotion recognition has attracted substantial research interest in human computer interfaces. However, the high dimensionality of EEG features, coupled with limited sample sizes, frequently leads to classifier overfitting and high computational complexity. Feature selection constitutes a critical strategy for mitigating these challenges. Most existing EEG feature selection methods assume complete multi-dimensional emotion labels. In practice, open acquisition environment, and the inherent subjectivity of emotion perception often result in incomplete label data, which can compromise model generalization. Additionally, existing feature selection methods for handling incomplete multi-dimensional labels primarily focus on correlations among various dimensions during label recovery, neglecting the correlation between samples in the label space and their interaction with various dimensions. To address these issues, we propose a novel incomplete multi-dimensional feature selection algorithm for EEG-based emotion recognition. The proposed method integrates an adaptive dual self-expression learning (ADSEL) with least squares regression. ADSEL establishes a bidirectional pathway between sample-level and dimension-level self-expression learning processes within the label space. It could facilitate the cross-sharing of learned information between these processes, enabling the simultaneous exploitation of effective information across both samples and dimensions for label reconstruction. Consequently, ADSEL could enhances label recovery accuracy and effectively identifies the optimal EEG feature subset for multi-dimensional emotion recognition.


Outsourced Privacy-Preserving Feature Selection Based on Fully Homomorphic Encryption

Wakiyama, Koki, I, Tomohiro, Sakamoto, Hiroshi

arXiv.org Artificial Intelligence

Feature selection is a technique that extracts a meaningful subset from a set of features in training data. When the training data is large-scale, appropriate feature selection enables the removal of redundant features, which can improve generalization performance, accelerate the training process, and enhance the interpretability of the model. This study proposes a privacy-preserving computation model for feature selection. Generally, when the data owner and analyst are the same, there is no need to conceal the private information. However, when they are different parties or when multiple owners exist, an appropriate privacy-preserving framework is required. Although various private feature selection algorithms, they all require two or more computing parties and do not guarantee security in environments where no external party can be fully trusted. To address this issue, we propose the first outsourcing algorithm for feature selection using fully homomorphic encryption. Compared to a prior two-party algorithm, our result improves the time and space complexity O(kn^2) to O(kn log^3 n) and O(kn), where k and n denote the number of features and data samples, respectively. We also implemented the proposed algorithm and conducted comparative experiments with the naive one. The experimental result shows the efficiency of our method even with small datasets.


Multi-label feature selection based on binary hashing learning and dynamic graph constraints

Guo, Cong, Huang, Changqin, Zhou, Wenhua, Huang, Xiaodi

arXiv.org Artificial Intelligence

Multi-label learning poses significant challenges in extracting reliable supervisory signals from the label space. Existing approaches often employ continuous pseudo-labels to replace binary labels, improving supervisory information representation. However, these methods can introduce noise from irrelevant labels and lead to unreliable graph structures. To overcome these limitations, this study introduces a novel multi-label feature selection method called Binary Hashing and Dynamic Graph Constraint (BHDG), the first method to integrate binary hashing into multi-label learning. BHDG utilizes low-dimensional binary hashing codes as pseudo-labels to reduce noise and improve representation robustness. A dynamically constrained sample projection space is constructed based on the graph structure of these binary pseudo-labels, enhancing the reliability of the dynamic graph. To further enhance pseudo-label quality, BHDG incorporates label graph constraints and inner product minimization within the sample space. Additionally, an $l_{2,1}$-norm regularization term is added to the objective function to facilitate the feature selection process. The augmented Lagrangian multiplier (ALM) method is employed to optimize binary variables effectively. Comprehensive experiments on 10 benchmark datasets demonstrate that BHDG outperforms ten state-of-the-art methods across six evaluation metrics. BHDG achieves the highest overall performance ranking, surpassing the next-best method by an average of at least 2.7 ranks per metric, underscoring its effectiveness and robustness in multi-label feature selection.


Metrics for Inter-Dataset Similarity with Example Applications in Synthetic Data and Feature Selection Evaluation -- Extended Version

Rajabinasab, Muhammad, Lautrup, Anton D., Zimek, Arthur

arXiv.org Artificial Intelligence

Measuring inter-dataset similarity is an important task in machine learning and data mining with various use cases and applications. Existing methods for measuring inter-dataset similarity are computationally expensive, limited, or sensitive to different entities and non-trivial choices for parameters. They also lack a holistic perspective on the entire dataset. In this paper, we propose two novel metrics for measuring inter-dataset similarity. We discuss the mathematical foundation and the theoretical basis of our proposed metrics. We demonstrate the effectiveness of the proposed metrics by investigating two applications in the evaluation of synthetic data and in the evaluation of feature selection methods. The theoretical and empirical studies conducted in this paper illustrate the effectiveness of the proposed metrics.